Embedding grammars into statistical language models

نویسندگان

  • Harald Hning
  • Manuel Kirschner
  • Fritz Class
  • André Berton
  • Udo Haiber
چکیده

This work combines grammars and statistical language models for speech recognition together in the same sentence. The grammars are compiled into bigrams with word indices, which serve to distinguish different syntactic positions of the same word. For both the grammatical and statistical parts there is one common interface for obtaining a language model score for bior trigrams. With only a small modification to a recogniser prepared for statistical language models, this new model can be applied without using a parser or a finite-state network in the recogniser. Priority is given to the grammar, therefore the combined model is able to disallow certain word transitions. With this combined language model, one or several grammatical phrases can be embedded into longer sentences.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Picture languages: Tiling systems versus tile rewriting grammars

Two formal models of pictures, i.e., 2D languages are compared: Tiling Systems and Tile Rewriting Grammars, which resp. extend to 2D the Regular and Context-Free languages. Two results extending classical language properties into 2D are proved. First, non-recursive TRG coincide with TS. Second, non-self-embedding TRG are suitably defined as corner grammars, showing that they generate TS languag...

متن کامل

Speech Recognition Grammar Compilation in Grammatical Framework

This paper describes how grammar-based language models for speech recognition systems can be generated from Grammatical Framework (GF) grammars. Context-free grammars and finite-state models can be generated in several formats: GSL, SRGS, JSGF, and HTK SLF. In addition, semantic interpretation code can be embedded in the generated context-free grammars. This enables rapid development of portabl...

متن کامل

Implicit Learning of Recursive Context-Free Grammars

Context-free grammars are fundamental for the description of linguistic syntax. However, most artificial grammar learning experiments have explored learning of simpler finite-state grammars, while studies exploring context-free grammars have not assessed awareness and implicitness. This paper explores the implicit learning of context-free grammars employing features of hierarchical organization...

متن کامل

Embedding attribute grammars and their extensions using functional zippers

Embedding Attribute Grammars and their Extensions using Functional Zippers Attribute grammars are a suitable formalism to express complex software language analysis and manipulation algorithms, which rely on multiple traversals of the underlying syntax tree. Attribute Grammars have been extended with mechanisms such as references, higher order and circular attributes. Such extensions provide a ...

متن کامل

Stochastic Categorial Grammars

Statistical methods have turned out to be quite successful in natural language processing. During the recent years, several models of stochastic grammars have been proposed, including models based on lexicalised context-free grammars [3], tree adjoining grammars [15], or dependency grammars [2, 5]. In this exploratory paper, we propose a new model of stochastic grammar, whose originality derive...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2005